-
Notifications
You must be signed in to change notification settings - Fork 296
Add support for float8 activation for Int4PreshuffledTensor #2437
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/2437
Note: Links to docs will display an error until the docs builds have been completed. ✅ You can merge normally! (1 Unrelated Failure)As of commit 889dca3 with merge base e5ca515 ( BROKEN TRUNK - The following job failed but were present on the merge base:👉 Rebase onto the `viable/strict` branch to avoid these failures
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
Summary: Note: slice is not working yet, others are working Test Plan: python test/dtypes/test_float8_activation_int4_groupwise_preshuffle.py Reviewers: Subscribers: Tasks: Tags: stack-info: PR: #2437, branch: jerryzh168/stack/4
d6d3477
to
26517e8
Compare
Summary: Note: slice is not working yet, others are working Test Plan: python test/dtypes/test_float8_activation_int4_groupwise_preshuffle.py Reviewers: Subscribers: Tasks: Tags: stack-info: PR: #2437, branch: jerryzh168/stack/4
26517e8
to
d187f78
Compare
d187f78
to
2fcff42
Compare
Summary: Note: slice is not working yet, others are working Test Plan: python test/dtypes/test_float8_activation_int4_groupwise_preshuffle.py Reviewers: Subscribers: Tasks: Tags: stack-info: PR: #2437, branch: jerryzh168/stack/4
Summary: Note: slice is not working yet, others are working Test Plan: python test/dtypes/test_float8_activation_int4_groupwise_preshuffle.py Reviewers: Subscribers: Tasks: Tags: stack-info: PR: #2437, branch: jerryzh168/stack/4
2fcff42
to
95856ed
Compare
Summary: Note: slice is not working yet, others are working Test Plan: python test/dtypes/test_float8_activation_int4_groupwise_preshuffle.py Reviewers: Subscribers: Tasks: Tags: stack-info: PR: #2437, branch: jerryzh168/stack/4
95856ed
to
1dec2cb
Compare
1dec2cb
to
1645c79
Compare
Summary: Note: slice is not working yet, others are working Test Plan: python test/dtypes/test_float8_activation_int4_groupwise_preshuffle.py Reviewers: Subscribers: Tasks: Tags: stack-info: PR: #2437, branch: jerryzh168/stack/4
Summary: Note: slice is not working yet, others are working Test Plan: python test/dtypes/test_float8_activation_int4_groupwise_preshuffle.py Reviewers: Subscribers: Tasks: Tags: stack-info: PR: #2437, branch: jerryzh168/stack/4
1645c79
to
5e9e869
Compare
72bc113
to
ea16397
Compare
test/quantization/quantize_/int4/test_int4_groupwise_preshuffle_tensor.py
Outdated
Show resolved
Hide resolved
test/quantization/quantize_/int4/test_int4_groupwise_preshuffle_tensor.py
Outdated
Show resolved
Hide resolved
torchao/quantization/quantize_/int4/int4_groupwise_preshuffle_tensor.py
Outdated
Show resolved
Hide resolved
torchao/quantization/quantize_/int4/int4_groupwise_preshuffle_tensor.py
Outdated
Show resolved
Hide resolved
torchao/quantization/quantize_/int4/int4_groupwise_preshuffle_tensor.py
Outdated
Show resolved
Hide resolved
torchao/quantization/quantize_/int4/int4_groupwise_preshuffle_tensor.py
Outdated
Show resolved
Hide resolved
torchao/quantization/quantize_/int4/int4_groupwise_preshuffle_tensor.py
Outdated
Show resolved
Hide resolved
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can you add serialization tests
you mean serialization of the models? it's moved to https://github.com/pytorch/ao/pull/2463/files#diff-9f6b6c4b39656e797cfda97536a4cf8a82004c64da518ad524637b471b716739, I don't exactly remember the reason for config serialization we did a config refactor in the last PR, I can add after we are aligned on what config should look like |
ea16397
to
c63ace9
Compare
c63ace9
to
ba6baf6
Compare
torchao/quantization/quantize_/int4/int4_groupwise_preshuffle_tensor.py
Outdated
Show resolved
Hide resolved
torchao/quantization/quantize_/int4/int4_groupwise_preshuffle_tensor.py
Outdated
Show resolved
Hide resolved
torchao/quantization/quantize_/int4/int4_groupwise_preshuffle_tensor.py
Outdated
Show resolved
Hide resolved
torchao/quantization/quantize_/int4/int4_groupwise_preshuffle_tensor.py
Outdated
Show resolved
Hide resolved
torchao/quantization/quantize_/int4/int4_groupwise_preshuffle_tensor.py
Outdated
Show resolved
Hide resolved
torchao/quantization/quantize_/int4/int4_groupwise_preshuffle_tensor.py
Outdated
Show resolved
Hide resolved
7dcfea8
to
96c3b6d
Compare
Summary: Added basic op support like linear and bmm, we have both float8 and bf16 in the same Tensor because it's the same dtype, only difference is whether the activation is quantized or not. Although there is some differneces in implementation: bf16 activaton: * group_scale * group_zero fp8 activation * group_scale * row_scale Test Plan: python test/quantization/quantize_/workflows/int4/test_int4_preshuffled_tensor.py Reviewers: Subscribers: Tasks: Tags: stack-info: PR: #2437, branch: jerryzh168/stack/4
96c3b6d
to
889dca3
Compare
Stacked PRs:
Add support for float8 activation for Int4PreshuffledTensor
Summary:
Added basic op support like linear and bmm, we have both float8 and bf16 in the same Tensor
because it's the same dtype, only difference is whether the activation is quantized or not. Although
there is some differneces in implementation:
bf16 activaton:
fp8 activation
Test Plan:
python test/quantization/quantize_/workflows/int4/test_int4_preshuffled_tensor.py
Reviewers:
Subscribers:
Tasks:
Tags: